Per-weight Class-based Learning Rates via Analytical Continuation
ثبت نشده
چکیده
We study the problem of training deep fully connected neural networks. Despite much progress in the design of activation functions, novel normalization techniques, and various skip-connection techniques, such networks remain challenging to train due to vanishing or exploding gradients. Our method is based on employing a different class-dependent learning rate to each network weight. Since the learning rates are hyperparameters and not part of the network, we perform an analytical continuation of the network, and create a generalized network. Following this reparameterization, the set of per-class per-weight learning rates are being manipulated during the training iterations. Our results show that the new algorithm leads to improved classification accuracy for both classical and modern activation functions.
منابع مشابه
A continuation approach for solving binary quadratic program based on a class of NCP-functions
In the paper, we consider a continuation approach for the binary quadratic program (BQP) based on a class of NCP-functions. More specifically, we recast the BQP as an equivalent minimization and then seeks its global minimizer via a global continuation method. Such approach had been considered in [11] which is based on the Fischer-Burmeister function. We investigate this continuation approach a...
متن کاملDiscriminative models for robust image classification
A variety of real-world tasks involve the classification of images into pre-determined categories. Designing image classification algorithms that exhibit robustness to acquisition noise and image distortions, particularly when the available training data are insufficient to learn accurate models, is a significant challenge. This dissertation explores the development of discriminative models for...
متن کاملThe Transforms of Fourier , the Transform of Laplas , the New for - Mula of Transformation
A new inverse formula for the Laplas's transformation. Pavlov An.V.(MIREA(TU)). In the article is proved,that the complex part of the analytical continuation of the r(p) = LLZ(x) = ∞ 0 e −pt dt ∞ 0 e −tx Z(x)dx, p ∈ {p : Im p ≥ 0}, equals to −πZ(x), x ∈ (0, ∞), if p = s = −x ∈ (−∞, 0) for a wide class of a functions Z(x) : It is proved,that the odd functions Z(x) = l k=1 γ k e λ k x , γ k = res...
متن کاملWeighted Instance-Based Learning Using Representative Intervals
Instance-based learning algorithms are widely used due to their capacity to approximate complex target functions; however, the performance of this kind of algorithms degrades significantly in the presence of irrelevant features. This paper introduces a new noise tolerant instance-based learning algorithm, called WIB-K, that uses one or more weights, per feature per class, to classify integer-va...
متن کاملIntrinsic Stabilization of Output Rates by Spike-Time Dependent Hebbian Learning
Over a broad parameter regime, spike-time dependent learning leads to an intrinsic stabilization of the mean firing rate of the postsynaptic neuron. Subtractive normalization of the synaptic weights (summed over all presynaptic inputs converging on a postsynaptic neuron) follows if, in addition, the mean input rates are identical at all synapses and correlations in the input are translation inv...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017